Attention-over-Attention Neural Networks for Reading Comprehension
نویسندگان
چکیده
Cloze-style queries are representative problems in reading comprehension. Over the past few months, we have seen much progress that utilizing neural network approach to solve Cloze-style questions. In this paper, we present a novel model called attention-over-attention reader for the Cloze-style reading comprehension task. Our model aims to place another attention mechanism over the document-level attention, and induces “attended attention” for final predictions. Unlike the previous works, our neural network model requires less pre-defined hyper-parameters and uses an elegant architecture for modeling. Experimental results show that the proposed attentionover-attention model significantly outperforms various state-of-the-art systems by a large margin in public datasets, such as CNN and Children’s Book Test datasets.
منابع مشابه
Consensus Attention-based Neural Networks for Chinese Reading Comprehension
Reading comprehension has embraced a booming in recent NLP research. Several institutes have released the Cloze-style reading comprehension data, and these have greatly accelerated the research of machine comprehension. In this work, we firstly present Chinese reading comprehension datasets, which consist of People Daily news dataset and Children’s Fairy Tale (CFT) dataset. Also, we propose a c...
متن کاملAssignment 4: Reading Comprehension
Reading comprehension is the task of understanding a piece of text by a machine. We train an end-to-end neural network that models the conditional distribution of start and end indices, given the question and context paragraph. We build on top of the baseline suggested in the Assignment, and explore new models to implement attention. We also measure the performance of the models and analyse the...
متن کاملCS224n Assignment 4: Machine Comprehension with Exploration on Attention Mechanism
This goal of this paper is to perform the prediction task on SQuAD dataset about reading comprehension. Given a pair of context paragraph and a question, we’ll output an answer. To do this, a model is built combining the idea of Bidirectional LSTM and attention flow mechanism. The basic architecture and setup details of the model are introduced, so do the summary of performance and error analys...
متن کاملIncreased resting-state functional connectivity of visual- and cognitive-control brain networks after training in children with reading difficulties
The Reading Acceleration Program, a computerized reading-training program, increases activation in neural circuits related to reading. We examined the effect of the training on the functional connectivity between independent components related to visual processing, executive functions, attention, memory, and language during rest after the training. Children 8-12 years old with reading difficult...
متن کاملWhy does working memory capacity predict variation in reading comprehension? On the influence of mind wandering and executive attention.
Some people are better readers than others, and this variation in comprehension ability is predicted by measures of working memory capacity (WMC). The primary goal of this study was to investigate the mediating role of mind-wandering experiences in the association between WMC and normal individual differences in reading comprehension, as predicted by the executive-attention theory of WMC (e.g.,...
متن کامل